Bayesian inverse reinforcement learning for collective animal movement

نویسندگان

چکیده

Agent-based methods allow for defining simple rules that generate complex group behaviors. The governing of such models are typically set a priori, and parameters tuned from observed behavior trajectories. Instead making simplifying assumptions across all anticipated scenarios, inverse reinforcement learning provides inference on the short-term (local) long-term policies by using properties Markov decision process. We use computationally efficient linearly-solvable process to learn local collective movement simulation selfpropelled-particle (SPP) model data application captive guppy population. estimation behavioral costs is done in Bayesian framework with basis function smoothing. recover true SPP find guppies value more than targeted toward shelter.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Bayesian Inverse Reinforcement Learning

Inverse Reinforcement Learning (IRL) is the problem of learning the reward function underlying a Markov Decision Process given the dynamics of the system and the behaviour of an expert. IRL is motivated by situations where knowledge of the rewards is a goal by itself (as in preference elicitation) and by the task of apprenticeship learning (learning policies from an expert). In this paper we sh...

متن کامل

Bayesian Multitask Inverse Reinforcement Learning

We generalise the problem of inverse reinforcement learning to multiple tasks, from multiple demonstrations. Each one may represent one expert trying to solve a different task, or as different experts trying to solve the same task. Our main contribution is to formalise the problem as statistical preference elicitation, via a number of structured priors, whose form captures our biases about the ...

متن کامل

Bayesian Nonparametric Inverse Reinforcement Learning

Inverse reinforcement learning (IRL) is the task of learning the reward function of a Markov Decision Process (MDP) given the transition function and a set of observed demonstrations in the form of stateaction pairs. Current IRL algorithms attempt to find a single reward function which explains the entire observation set. In practice, this leads to a computationally-costly search over a large (...

متن کامل

MAP Inference for Bayesian Inverse Reinforcement Learning

The difficulty in inverse reinforcement learning (IRL) arises in choosing the best reward function since there are typically an infinite number of reward functions that yield the given behaviour data as optimal. Using a Bayesian framework, we address this challenge by using the maximum a posteriori (MAP) estimation for the reward function, and show that most of the previous IRL algorithms can b...

متن کامل

Bayesian Nonparametric Feature Construction for Inverse Reinforcement Learning

Most of the algorithms for inverse reinforcement learning (IRL) assume that the reward function is a linear function of the pre-defined state and action features. However, it is often difficult to manually specify the set of features that can make the true reward function representable as a linear function. We propose a Bayesian nonparametric approach to identifying useful composite features fo...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: The Annals of Applied Statistics

سال: 2022

ISSN: ['1941-7330', '1932-6157']

DOI: https://doi.org/10.1214/21-aoas1529